2025-04-06 19:32:29,245 [ 66879 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-06 19:32:29,245 [ 66879 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-06 19:32:29,245 [ 66879 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-06 19:32:29,245 [ 66879 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_e8qakn --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: random-0.2, timeout-2.2.0, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_accept_invalid_certificate/test.py::test_accept test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] test_backup_restore/test.py::test_attach_partition test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_default_role/test.py::test_alter_user test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_cluster_all_replicas/test.py::test_cluster test_attach_partition_using_copy/test.py::test_all_replicated test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_config_decryption/test_wrong_settings.py::test_invalid_chars [gw7] [ 1%] PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key [gw4] [ 2%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] [gw4] [ 3%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] [gw3] [ 4%] PASSED test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept [gw3] [ 5%] PASSED test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default [gw3] [ 6%] PASSED test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject [gw3] [ 7%] PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject [gw4] [ 8%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] [gw3] [ 9%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw3] [ 10%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw4] [ 11%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] [gw9] [ 12%] PASSED test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles [gw4] [ 13%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] [gw9] [ 14%] PASSED test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role [gw4] [ 15%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] test_concurrent_queries_restriction_by_query_kind/test.py::test_insert [gw9] [ 16%] PASSED test_default_role/test.py::test_wrong_set_default_role [gw4] [ 17%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] [gw8] [ 18%] PASSED test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition [gw0] [ 19%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_create_query_constraints/test.py::test_create_query_const_constraints [gw5] [ 20%] PASSED test_cluster_all_replicas/test.py::test_cluster test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] [gw4] [ 21%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] [gw7] [ 22%] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_custom_settings/test.py::test_custom_settings [gw0] [ 23%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] [gw5] [ 24%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] [gw8] [ 25%] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] [gw1] [ 26%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] [gw0] [ 27%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] [gw7] [ 28%] PASSED test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method [gw5] [ 29%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] test_cluster_all_replicas/test.py::test_global_in [gw8] [ 30%] PASSED test_backup_restore/test.py::test_restore test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica [gw5] [ 31%] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] [gw3] [ 32%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select [gw4] [ 33%] PASSED test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting [gw4] [ 34%] PASSED test_custom_settings/test.py::test_illformed_setting [gw9] [ 35%] PASSED test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] test_always_fetch_merged/test.py::test_replica_always_download [gw0] [ 36%] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw9] [ 37%] PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes [gw5] [ 38%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] [gw0] [ 39%] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_async_insert_memory/test.py::test_memory_usage test_backup_restore_on_cluster/test.py::test_empty_replicated_table [gw7] [ 40%] PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field [gw0] [ 41%] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table [gw5] [ 42%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] test_backup_restore_on_cluster/test.py::test_file_deduplication [gw2] [ 43%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] [gw0] [ 44%] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication [gw1] [ 45%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] test_backup_restore_on_cluster/test.py::test_get_error_from_other_host [gw0] [ 46%] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host [gw3] [ 47%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_config_yaml_main/test.py::test_yaml_main_conf test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] [gw0] [ 48%] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation [gw8] [ 49%] PASSED test_always_fetch_merged/test.py::test_replica_always_download [gw7] [ 50%] PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field [gw0] [ 51%] PASSED test_backup_restore_on_cluster/test.py::test_mutation test_compression_nested_columns/test.py::test_nested_compression_codec test_config_xml_main/test.py::test_xml_main_conf test_backup_restore_on_cluster/test.py::test_projection [gw1] [ 52%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] [gw0] [ 53%] PASSED test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] [gw5] [ 54%] PASSED test_config_yaml_main/test.py::test_yaml_main_conf [gw0] [ 55%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async [gw1] [ 56%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] [gw2] [ 57%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] [gw7] [ 58%] PASSED test_config_xml_main/test.py::test_xml_main_conf [gw8] [ 59%] PASSED test_compression_nested_columns/test.py::test_nested_compression_codec [gw0] [ 60%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 61%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] [gw1] [ 62%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] [gw0] [ 63%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] [gw3] [ 64%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] [gw0] [ 65%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] test_backup_restore_on_cluster/test.py::test_replicated_table [gw0] [ 66%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table [gw1] [ 67%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster [gw8] [ 68%] PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 69%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] [gw2] [ 70%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster [gw0] [ 71%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def [gw0] [ 72%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert [gw2] [ 73%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] [gw0] [ 74%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert [gw1] [ 75%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge [gw0] [ 76%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath [gw3] [ 77%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] [gw0] [ 78%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges [gw0] [ 79%] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup [gw1] [ 80%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] test_dictionaries_dependency/test.py::test_no_lazy_load [gw2] [ 81%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] [gw0] [ 82%] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions [gw0] [ 83%] PASSED test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users [gw0] [ 84%] PASSED test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def [gw0] [ 85%] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty [gw0] [ 86%] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency [gw4] [ 87%] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw0] [ 88%] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency [gw1] [ 89%] PASSED test_dictionaries_dependency/test.py::test_no_lazy_load [gw2] [ 90%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica [gw9] [ 91%] PASSED test_async_insert_memory/test.py::test_memory_usage test_ddl_worker_replicas/test.py::test_ddl_worker_replicas [gw4] [ 92%] PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica [gw2] [ 93%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] test_disk_access_storage/test.py::test_alter [gw2] [ 94%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] [gw4] [ 95%] PASSED test_disk_access_storage/test.py::test_alter [gw9] [ 96%] PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas [gw6] [ 97%] FAILED test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree [gw6] [ 98%] FAILED test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk [gw6] [ 99%] FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated [gw6] [100%] FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated =================================== FAILURES =================================== _____________________________ test_all_replicated ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:126: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ---------------------------- Captured stderr setup ----------------------------- Command:[docker ps | wc -l] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py Cluster start called. is_up=False Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: replica1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: replica2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] Stderr: replica2 Skipped - Image is already being pulled by zoo3 Stderr: replica1 Skipped - Image is already being pulled by zoo3 Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr:time="2025-04-06T19:32:46Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating Stderr: Network roottestattachpartitionusingcopy-gw6_default Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started Stderr:time="2025-04-06T19:32:47Z" level=debug msg="otel error" error="" Stderr:time="2025-04-06T19:32:47Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.3.2, port:2181, use_ssl:False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.3.3, port:2181, use_ssl:False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.3.4, port:2181, use_ssl:False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started ClickHouse instance created get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica1, ip: 172.16.3.6... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None ClickHouse replica1 started get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica2, ip: 172.16.3.5... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f7aa99407ab3c9e154aab78c47f499d7d78d831625b296c5a42a8562adc5f9ad/json HTTP/1.1" 200 None ClickHouse replica2 started ------------------------------ Captured log setup ------------------------------ 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : No running containers (conftest.py:96, cleanup_environment) 2025-04-06 19:32:35 [ 667 ] DEBUG : Pruning Docker networks (conftest.py:98, cleanup_environment) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker network prune --force] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:146, run_and_check) 2025-04-06 19:32:35 [ 667 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2793, start) 2025-04-06 19:32:35 [ 667 ] DEBUG : Cluster start called. is_up=False (cluster.py:2800, start) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-06 19:32:35 [ 667 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2813, start) 2025-04-06 19:32:35 [ 667 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d (cluster.py:4752, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database (cluster.py:4769, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs (cluster.py:4780, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2813, start) 2025-04-06 19:32:35 [ 667 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d (cluster.py:4752, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database (cluster.py:4769, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs (cluster.py:4780, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-06 19:32:35 [ 667 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env (cluster.py:97, _create_env_file) 2025-04-06 19:32:35 [ 667 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-06 19:32:35 [ 667 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-06 19:32:35 [ 667 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-06 19:32:35 [ 667 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-06 19:32:35 [ 667 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-06 19:32:35 [ 667 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] (cluster.py:122, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: replica1 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: zoo3 Pulling (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Stderr: zoo3 Pulled (cluster.py:148, run_and_check) 2025-04-06 19:32:46 [ 667 ] DEBUG : Setup ZooKeeper (cluster.py:2854, start) 2025-04-06 19:32:46 [ 667 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] (cluster.py:2855, start) 2025-04-06 19:32:46 [ 667 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:122, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr:time="2025-04-06T19:32:46Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Created (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr:time="2025-04-06T19:32:47Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Stderr:time="2025-04-06T19:32:47Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-06 19:32:47 [ 667 ] DEBUG : Wait ZooKeeper to start (cluster.py:2466, wait_zookeeper_to_start) 2025-04-06 19:32:47 [ 667 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2082, get_instance_ip) 2025-04-06 19:32:47 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:47 [ 667 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.3.2, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-06 19:32:47 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:47 [ 667 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-06 19:32:47 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:47 [ 667 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-06 19:32:47 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:47 [ 667 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-06 19:32:48 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:48 [ 667 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-06 19:32:49 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:49 [ 667 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-06 19:32:51 [ 667 ] INFO : Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:51 [ 667 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-06 19:32:51 [ 667 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-06 19:32:51 [ 667 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-06 19:32:51 [ 667 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-06 19:32:51 [ 667 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-06 19:32:51 [ 667 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-06 19:32:51 [ 667 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-06 19:32:51 [ 667 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-06 19:32:52 [ 667 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2082, get_instance_ip) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.3.3, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-06 19:32:52 [ 667 ] INFO : Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-06 19:32:52 [ 667 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-06 19:32:52 [ 667 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2082, get_instance_ip) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.3.4, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-06 19:32:52 [ 667 ] INFO : Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-06 19:32:52 [ 667 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-06 19:32:52 [ 667 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-06 19:32:52 [ 667 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-06 19:32:52 [ 667 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-06 19:32:52 [ 667 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-06 19:32:52 [ 667 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2482, wait_zookeeper_nodes_to_start) 2025-04-06 19:32:52 [ 667 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3200, start) 2025-04-06 19:32:52 [ 667 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:122, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started (cluster.py:148, run_and_check) 2025-04-06 19:32:52 [ 667 ] DEBUG : ClickHouse instance created (cluster.py:3208, start) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2082, get_instance_ip) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2092, get_instance_global_ipv6) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.3.6... (cluster.py:3216, start) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:52 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:53 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/5252ddc5d0c4e8e3ca125a6a86e4eb2507410349788a11e88306e1ec17331938/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : ClickHouse replica1 started (cluster.py:3220, start) 2025-04-06 19:32:54 [ 667 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2082, get_instance_ip) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2092, get_instance_global_ipv6) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.3.5... (cluster.py:3216, start) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : http://localhost:None "GET /v1.46/containers/f7aa99407ab3c9e154aab78c47f499d7d78d831625b296c5a42a8562adc5f9ad/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-06 19:32:54 [ 667 ] DEBUG : ClickHouse replica2 started (cluster.py:3220, start) ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (140467834857024) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-06 19:32:54 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-06 19:32:55 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-06 19:32:55 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-06 19:32:55 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-06 19:32:55 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:33:50 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:34:44 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:35:39 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:36:36 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:37:34 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:38:29 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:39:24 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:40:21 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:41:19 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:42:13 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:43:09 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:44:06 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:45:03 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:45:58 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:46:53 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _____________________________ test_both_mergetree ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:104: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (140467834857024) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-06 19:47:35 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-06 19:47:50 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-06 19:47:50 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-06 19:47:50 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-06 19:47:51 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:48:48 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:49:43 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:50:38 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:51:35 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:52:33 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:53:28 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:54:23 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:55:20 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:56:18 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:57:12 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:58:08 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 19:59:05 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:00:02 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:00:57 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:01:52 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_not_work_on_different_disk ________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:197: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (140467834857024) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-06 20:02:35 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-06 20:02:49 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-06 20:02:49 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-06 20:02:49 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-06 20:02:49 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:03:47 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:04:42 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:05:37 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:06:34 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:07:32 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:08:27 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:09:22 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:10:19 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:11:17 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:12:11 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:13:06 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:14:03 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:15:01 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:15:56 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:16:51 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_only_destination_replicated _______________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (140467834857024) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-06 20:17:35 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-06 20:17:48 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-06 20:17:48 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-06 20:17:48 [ 667 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-06 20:17:48 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:18:46 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:19:41 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:20:36 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:21:33 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:22:31 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:23:25 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:24:21 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:25:18 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:26:15 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:27:10 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:28:05 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:29:02 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:30:00 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:30:55 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-06 20:31:50 [ 667 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) --------------------------- Captured stderr teardown --------------------------- Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ---------------------------- Captured log teardown ----------------------------- 2025-04-06 20:32:36 [ 667 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] (cluster.py:122, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-06 20:32:42 [ 667 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] (cluster.py:122, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed (cluster.py:148, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-06 20:32:43 [ 667 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-06 20:32:43 [ 667 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-06 20:32:43 [ 667 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-06 20:32:43 [ 667 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-06 20:32:43 [ 667 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) ============================== slowest durations =============================== 900.00s call test_attach_partition_using_copy/test.py::test_both_mergetree 900.00s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 900.00s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 880.33s call test_attach_partition_using_copy/test.py::test_all_replicated 187.94s call test_async_insert_memory/test.py::test_memory_usage 176.65s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 86.03s call test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 50.51s call test_dictionaries_dependency/test.py::test_no_lazy_load 42.19s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 41.92s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 41.72s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 41.07s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 41.04s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 39.70s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 39.14s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 38.52s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 37.75s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 25.71s setup test_cluster_all_replicas/test.py::test_cluster 23.35s setup test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 22.76s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 21.91s teardown test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 21.29s call test_disk_access_storage/test.py::test_alter 21.28s call test_concurrent_queries_restriction_by_query_kind/test.py::test_select 20.80s call test_config_yaml_main/test.py::test_yaml_main_conf 19.68s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 19.68s setup test_attach_partition_using_copy/test.py::test_all_replicated 19.30s call test_config_xml_main/test.py::test_xml_main_conf 19.20s setup test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 18.96s setup test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 18.87s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 18.48s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 18.41s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 18.39s setup test_compression_nested_columns/test.py::test_nested_compression_codec 17.97s setup test_backup_restore/test.py::test_attach_partition 17.92s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 17.58s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 17.54s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 16.30s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 16.09s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 16.07s setup test_always_fetch_merged/test.py::test_replica_always_download 15.70s setup test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 15.46s setup test_accept_invalid_certificate/test.py::test_accept 14.94s setup test_disk_access_storage/test.py::test_alter 14.88s setup test_default_role/test.py::test_alter_user 14.79s call test_config_decryption/test_wrong_settings.py::test_no_encryption_key 14.41s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 14.28s call test_config_decryption/test_wrong_settings.py::test_invalid_chars 14.13s teardown test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 13.68s setup test_create_query_constraints/test.py::test_create_query_const_constraints 13.34s call test_config_decryption/test_wrong_settings.py::test_wrong_method 13.34s setup test_async_insert_memory/test.py::test_memory_usage 12.85s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 12.82s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 11.49s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 11.47s call test_always_fetch_merged/test.py::test_replica_always_download 10.63s call test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 10.54s call test_config_decryption/test_wrong_settings.py::test_subnodes 10.32s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 10.02s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 9.91s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 9.80s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 8.38s call test_backup_restore/test.py::test_replace_partition 8.33s setup test_custom_settings/test.py::test_custom_settings 8.25s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 8.22s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 8.13s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 8.02s teardown test_dictionaries_dependency/test.py::test_no_lazy_load 7.67s call test_backup_restore/test.py::test_attach_partition 7.50s call test_backup_restore_on_cluster/test.py::test_tables_dependency 7.44s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 6.76s call test_backup_restore_on_cluster/test.py::test_required_privileges 6.38s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 6.38s call test_backup_restore_on_cluster/test.py::test_replicated_database 6.36s call test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 6.30s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 6.29s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 6.29s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.16s call test_backup_restore/test.py::test_restore 6.06s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 6.02s call test_backup_restore_on_cluster/test.py::test_mutation 5.91s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 5.90s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 5.88s teardown test_always_fetch_merged/test.py::test_replica_always_download 5.75s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 5.70s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 5.69s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 5.66s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 5.66s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 5.61s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 5.60s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 5.50s teardown test_compression_nested_columns/test.py::test_nested_compression_codec 5.33s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 5.15s teardown test_accept_invalid_certificate/test.py::test_strict_reject_with_config 5.07s teardown test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 5.05s teardown test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 4.94s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 4.85s teardown test_async_insert_memory/test.py::test_memory_usage 4.81s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 4.72s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 4.71s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 4.64s call test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 4.57s teardown test_create_query_constraints/test.py::test_create_query_minmax_constraints 4.53s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_select 4.36s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 4.34s call test_backup_restore_on_cluster/test.py::test_system_functions 4.32s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 4.28s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 4.25s call test_backup_restore_on_cluster/test.py::test_replicated_table 4.23s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 4.23s call test_backup_restore_on_cluster/test.py::test_system_users 4.20s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 4.20s call test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 4.09s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 3.83s call test_backup_restore_on_cluster/test.py::test_projection 3.81s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 3.75s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 3.74s teardown test_default_role/test.py::test_wrong_set_default_role 3.68s call test_backup_restore_on_cluster/test.py::test_file_deduplication 3.58s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 3.56s teardown test_backup_restore/test.py::test_restore 3.45s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 3.32s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 3.31s teardown test_custom_settings/test.py::test_illformed_setting 3.31s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 3.24s call test_create_query_constraints/test.py::test_create_query_minmax_constraints 3.20s call test_default_role/test.py::test_set_default_roles 3.02s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 3.00s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 2.99s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 2.97s call test_compression_nested_columns/test.py::test_nested_compression_codec 2.81s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 2.73s call test_default_role/test.py::test_alter_user 2.64s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 2.56s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 2.45s call test_create_query_constraints/test.py::test_create_query_const_constraints 2.08s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 2.03s teardown test_backup_restore_on_cluster/test.py::test_mutation 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 1.88s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.78s teardown test_backup_restore_on_cluster/test.py::test_projection 1.78s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 1.73s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.73s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 1.73s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.69s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 1.68s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.68s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 1.68s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 1.68s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 1.68s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 1.63s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 1.63s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.63s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 1.63s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.63s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 1.63s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 1.63s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.58s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.58s teardown test_backup_restore_on_cluster/test.py::test_system_functions 1.58s teardown test_backup_restore_on_cluster/test.py::test_system_users 1.53s call test_cluster_all_replicas/test.py::test_cluster 1.45s call test_custom_settings/test.py::test_custom_settings 1.34s call test_default_role/test.py::test_wrong_set_default_role 1.26s teardown test_disk_access_storage/test.py::test_alter 1.10s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.80s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.75s call test_cluster_all_replicas/test.py::test_global_in 0.75s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.70s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.70s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.66s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.49s call test_custom_settings/test.py::test_illformed_setting 0.32s setup test_default_role/test.py::test_wrong_set_default_role 0.32s call test_accept_invalid_certificate/test.py::test_connection_accept 0.32s call test_accept_invalid_certificate/test.py::test_accept 0.22s setup test_default_role/test.py::test_set_default_roles 0.22s call test_accept_invalid_certificate/test.py::test_strict_reject 0.22s call test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.22s call test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.18s call test_accept_invalid_certificate/test.py::test_default 0.04s setup test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s teardown test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 0.00s setup test_create_query_constraints/test.py::test_create_query_minmax_constraints 0.00s setup test_backup_restore/test.py::test_restore 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_select 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s teardown test_config_yaml_main/test.py::test_yaml_main_conf 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s teardown test_default_role/test.py::test_alter_user 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 0.00s teardown test_create_query_constraints/test.py::test_create_query_const_constraints 0.00s setup test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s teardown test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_dictionaries_dependency/test.py::test_no_lazy_load 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 0.00s setup test_custom_settings/test.py::test_illformed_setting 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s setup test_config_yaml_main/test.py::test_yaml_main_conf 0.00s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s setup test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 0.00s teardown test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.00s teardown test_accept_invalid_certificate/test.py::test_accept 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s setup test_accept_invalid_certificate/test.py::test_connection_accept 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s setup test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_custom_settings/test.py::test_custom_settings 0.00s teardown test_accept_invalid_certificate/test.py::test_default 0.00s setup test_accept_invalid_certificate/test.py::test_default 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s teardown test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s teardown test_default_role/test.py::test_set_default_roles 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated 0.00s setup test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_reject 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_accept_invalid_certificate/test.py::test_connection_accept =========================== short test summary info ============================ FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Failed... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Failed... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] PASSED test_accept_invalid_certificate/test.py::test_accept PASSED test_accept_invalid_certificate/test.py::test_connection_accept PASSED test_accept_invalid_certificate/test.py::test_default PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] PASSED test_accept_invalid_certificate/test.py::test_strict_reject PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] PASSED test_default_role/test.py::test_alter_user PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] PASSED test_default_role/test.py::test_set_default_roles PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] PASSED test_default_role/test.py::test_wrong_set_default_role PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] PASSED test_backup_restore/test.py::test_attach_partition PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore/test.py::test_replace_partition PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_config_decryption/test_wrong_settings.py::test_subnodes PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_backup_restore/test.py::test_restore PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert PASSED test_custom_settings/test.py::test_custom_settings PASSED test_custom_settings/test.py::test_illformed_setting PASSED test_create_query_constraints/test.py::test_create_query_const_constraints PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_always_fetch_merged/test.py::test_replica_always_download PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_config_yaml_main/test.py::test_yaml_main_conf PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] PASSED test_config_xml_main/test.py::test_xml_main_conf PASSED test_compression_nested_columns/test.py::test_nested_compression_codec PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_dictionaries_dependency/test.py::test_no_lazy_load PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] PASSED test_async_insert_memory/test.py::test_memory_usage PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] PASSED test_disk_access_storage/test.py::test_alter PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas ================== 4 failed, 96 passed in 3610.15s (1:00:10) =================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 528, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_e8qakn --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca ' returned non-zero exit status 1.